Skip to main content

About the Provider

Alibaba Cloud is the cloud computing arm of Alibaba Group and the creator of the Qwen model family. Through its open-source initiative, Alibaba has released state-of-the-art language and multimodal models under permissive licenses, enabling developers and enterprises to build powerful AI applications across diverse domains and languages.

Model Quickstart

This section helps you quickly get started with the Qwen/Qwen3-Plus model on the Qubrid AI inferencing platform. To use this model, you need:
  • A valid Qubrid API key
  • Access to the Qubrid inference API
  • Basic knowledge of making API requests in your preferred language
Once authenticated with your API key, you can send inference requests to the Qwen/Qwen3-Plus model and receive responses based on your input prompts. Below are example placeholders showing how the model can be accessed using different programming environments.
You can choose the one that best fits your workflow.
from openai import OpenAI

# Initialize the OpenAI client with Qubrid base URL
client = OpenAI(
    base_url="https://platform.qubrid.com/v1",
    api_key="QUBRID_API_KEY",
)

# Create a streaming chat completion
stream = client.chat.completions.create(
    model="Qwen/Qwen3-Plus",
    messages=[
      {
        "role": "user",
        "content": "Explain quantum computing in simple terms"
      }
    ],
    max_tokens=4096,
    temperature=0.7,
    top_p=1,
    stream=True
)

# If stream = False comment this out
for chunk in stream:
    if chunk.choices and chunk.choices[0].delta.content:
        print(chunk.choices[0].delta.content, end="", flush=True)
print("\n")

# If stream = True comment this out
print(stream.choices[0].message.content)

Model Overview

Qwen3 Plus is a balanced general-purpose model for everyday chat and analysis tasks.
  • Built on a Transformer decoder-only architecture with up to 1M token context, it delivers fast and reliable responses across a wide range of use cases.
  • With strong multilingual support and instruction tuning on multilingual web data, it is well suited for customer support, business writing, and ideation workflows.

Model at a Glance

FeatureDetails
Model IDQwen/Qwen3-Plus
ProviderAlibaba Cloud (Qwen Team)
ArchitectureTransformer decoder-only
Model SizeN/A
Parameters4
Context LengthUp to 1M Tokens
Release Date2025
LicenseApache 2.0
Training DataMultilingual web data with instruction tuning

When to use?

You should consider using Qwen3 Plus if:
  • You need customer support chatbots that resolve common issues and FAQs across multiple languages
  • Your application requires business and marketing writing such as emails, blog posts, social copy, and internal docs
  • You are building brainstorming and ideation workflows for product features, campaign concepts, and content outlines

Inference Parameters

Parameter NameTypeDefaultDescription
StreamingbooleantrueEnable streaming responses for real-time output.
Temperaturenumber0.7Controls randomness. Higher values mean more creative but less predictable output.
Max Tokensnumber4096Maximum number of tokens to generate in the response.
Top Pnumber1Nucleus sampling: considers tokens with top_p probability mass.

Key Features

  • Fast and Reliable: Optimized for low-latency responses across everyday chat and analysis tasks.
  • Strong Multilingual Support: Trained on multilingual web data for consistent performance across languages.
  • Up to 1M Token Context: Supports long conversations, documents, and extended workflows.
  • Apache 2.0 License: Fully open-source with unrestricted commercial use.

Summary

Qwen3 Plus is Alibaba’s balanced general-purpose chat model built for speed, reliability, and multilingual coverage.
  • It uses a Transformer decoder-only architecture with up to 1M token context and instruction tuning on multilingual web data.
  • It is optimized for customer support, business writing, and ideation workflows at low cost.
  • The model delivers fast, reliable responses across multiple languages with a 1M token context window.
  • Licensed under Apache 2.0 for full commercial use.